Learning in a Large Function Space: Privacy-Preserving Mechanisms for SVM Learning
نویسندگان
چکیده
Several recent studies in privacy-preserving learning have considered the trade-o between utility or risk and the level of di erential privacy guaranteed by mechanisms for statistical query processing. In this paper we study this trade-o in private Support Vector Machine (SVM) learning. We present two e cient mechanisms, one for the case of nite-dimensional feature mappings and one for potentially in nite-dimensional feature mappings with translation-invariant kernels. For the case of translation-invariant kernels, the proposed mechanism minimizes regularized empirical risk in a random Reproducing Kernel Hilbert Space whose kernel uniformly approximates the desired kernel with high probability. This technique, borrowed from large-scale learning, allows the mechanism to respond with a nite encoding of the classi er, even when the function class is of in nite VC dimension. Di erential privacy is established using a proof technique from algorithmic stability. Utility the mechanism's response function is pointwise -close to non-private SVM with probability 1− δ is proven by appealing to the smoothness of regularized empirical risk minimization with respect to small perturbations to the feature mapping. We conclude with a lower bound on the optimal di erential privacy of the SVM. This negative result states that for any δ, no mechanism can be simultaneously ( , δ)-useful and β-di erentially private for small and small β.
منابع مشابه
Privacy Preserving Machine Learning: Related Work
A practical scenario of PPML is where only one central party has the entire data on which the ML algorithm has to be learned. Agrawal and Ramakrishnan [1] proposed the first method to learn a Decision Tree classifier on a database without revealing any information about individual records. They consider public model private data setting where the algorithm and its parameters are public whereas ...
متن کاملStatement of Research — Alexandre Evfimievski
My prior research has been mainly in the area of privacy preserving data mining. It included such topics as: using randomization for preserving privacy of individual transactions in association rule mining; secure two-party computation of joins between two relational tables, set intersections, join sizes, and supports of vertically partitioned itemsets; improving space and time efficiency in pr...
متن کاملThe Large Margin Mechanism for Differentially Private Maximization
A basic problem in the design of privacy-preserving algorithms is the private maximization problem: the goal is to pick an item from a universe that (approximately) maximizes a data-dependent function, all under the constraint of differential privacy. This problem has been used as a sub-routine in many privacy-preserving algorithms for statistics and machine-learning. Previous algorithms for th...
متن کاملDynamic Privacy For Distributed Machine Learning Over Network
Privacy-preserving distributed machine learning becomes increasingly important due to the recent rapid growth of data. This paper focuses on a class of regularized empirical risk minimization (ERM) machine learning problems, and develops two methods to provide differential privacy to distributed learning algorithms over a network. We first decentralize the learning algorithm using the alternati...
متن کاملShare your Model instead of your Data: Privacy Preserving Mimic Learning for Ranking
Deep neural networks have become a primary tool for solving problems in many elds. ey are also used for addressing information retrieval problems and show strong performance in several tasks. Training these models requires large, representative datasets and for most IR tasks, such data contains sensitive information from users. Privacy and condentiality concerns prevent many data owners from...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/0911.5708 شماره
صفحات -
تاریخ انتشار 2009